Goto

Collaborating Authors

 multi-class classification


A Generalised Exponentiated Gradient Approach to Enhance Fairness in Binary and Multi-class Classification Tasks

Boubekraoui, Maryam, d'Aloisio, Giordano, Di Marco, Antinisca

arXiv.org Machine Learning

The widespread use of AI and ML models in sensitive areas raises significant concerns about fairness. While the research community has introduced various methods for bias mitigation in binary classification tasks, the issue remains under-explored in multi-class classification settings. To address this limitation, in this paper, we first formulate the problem of fair learning in multi-class classification as a multi-objective problem between effectiveness (i.e., prediction correctness) and multiple linear fairness constraints. Next, we propose a Generalised Exponentiated Gradient (GEG) algorithm to solve this task. GEG is an in-processing algorithm that enhances fairness in binary and multi-class classification settings under multiple fairness definitions. We conduct an extensive empirical evaluation of GEG against six baselines across seven multi-class and three binary datasets, using four widely adopted effectiveness metrics and three fairness definitions. GEG overcomes existing baselines, with fairness improvements up to 92% and a decrease in accuracy up to 14%.



Multi-Class Learning: From Theory to Algorithm

Jian Li, Yong Liu, Rong Yin, Hua Zhang, Lizhong Ding, Weiping Wang

Neural Information Processing Systems

Moreover,the proposed multi-class kernel learning algorithms have statistical guarantees and fast convergence rates. Experimental results on lots of benchmark datasets show that our proposed methods can significantly outperform the existing multi-class classification methods. The major contributions ofthispaper include: 1)Anewlocal Rademacher complexitybased bound withfastconvergence rate for multi-class classification is established. Existing works [16,27] for multi-class classifiers with Rademacher complexity does not take into account couplings among different classes.




SupplementaryMaterial CanLessbeMore? WhenIncreasing-to-Balancing LabelNoiseRatesConsideredBeneficial

Neural Information Processing Systems

A.10 Extensiontomulti-class As explained at the beginning, our algorithm can largely extend to the multi-class/group setting. The task is to predict whether an individual's income exceeds50K.Thedatasetconsists of48,842examples and28features. Fairface, the face attribute dataset containing 108,501 images with balanced race and gender groups [15]. We use a pre-trained vision transformer (ViT/B-32) model [8] to extract image representations, and project them into 50-dimensional feature vectors. For constrained learning, we categorize race into White and Non-White groups.



H-Consistency Bounds: Characterization and Extensions

Neural Information Processing Systems

These are upper bounds on the zero-one estimation error of any predictor in a hypothesis set, expressed in terms of its surrogate loss estimation error.



Demystifying the Optimal Performance of Multi-Class Classification

Neural Information Processing Systems

Classification is a fundamental task in science and engineering on which machine learning methods have shown outstanding performances. However, it is challenging to determine whether such methods have achieved the Bayes error rate, that is, the lowest error rate attained by any classifier. This is mainly due to the fact that the Bayes error rate is not known in general and hence, effectively estimating it is paramount. Inspired by the work by Ishida et al. (2023), we propose an estimator for the Bayes error rate of supervised multi-class classification problems. We analyze several theoretical aspects of such estimator, including its consistency, unbiasedness, convergence rate, variance, and robustness. We also propose a denoising method that reduces the noise that potentially corrupts the data labels, and we improve the robustness of the proposed estimator to outliers by incorporating the median-of-means estimator. Our analysis demonstrates the consistency, asymptotic unbiasedness, convergence rate, and robustness of the proposed estimators.